Goto

Collaborating Authors

 Lâm Đồng Province


PRKAN: Parameter-Reduced Kolmogorov-Arnold Networks

Ta, Hoang-Thang, Thai, Duy-Quy, Tran, Anh, Sidorov, Grigori, Gelbukh, Alexander

arXiv.org Artificial Intelligence

MLPs have been one of key components in modern neural network architectures for years. Their simplicity makes them widely used for capturing complex relationships through multiple layers of non-linear transformations. However, their role has been reconsidered recently with the revival of Kolmogorov-Arnold Networks (KANs) [1, 2]. In these papers, fixed activation functions used in MLPs are described as "nodes," and the authors proposed replacing them with learnable activation functions like B-splines, referred to as "edges", to improve performance in mathematical and physical examples. To address Hilbert's 13th problem [3], the Kolmogorov-Arnold Representation Theorem (KART) [4] was introduced. It posits that any continuous function involving multiple variables can be decomposed into a sum of continuous functions of single variables, thus inspiring the creation of KANs. The work of Liu et al. [1] on KANs has inspired numerous studies exploring the use of various basis and polynomial functions as replacements for B-splines [5, 6, 7, 8, 9, 10, 11, 12, 13], investigating the model's performance compared to MLPs. Several studies have shown that KANs do not always outperform MLPs when using the same training parameters [14, 15]. Moreover, while KANs achieve better performance than MLPs with the same network structure, they often require a significantly larger number of parameters [7, 16, 17, 18, 19].


Evaluating Large Language Model Capability in Vietnamese Fact-Checking Data Generation

To, Long Truong, Le, Hung Tuan, Nguyen, Dat Van-Thanh, Nguyen, Manh Trong, Nguyen, Tri Thien, Van Huynh, Tin, Van Nguyen, Kiet

arXiv.org Artificial Intelligence

Large Language Models (LLMs), with gradually improving reading comprehension and reasoning capabilities, are being applied to a range of complex language tasks, including the automatic generation of language data for various purposes. However, research on applying LLMs for automatic data generation in low-resource languages like Vietnamese is still underdeveloped and lacks comprehensive evaluation. In this paper, we explore the use of LLMs for automatic data generation for the Vietnamese fact-checking task, which faces significant data limitations. Specifically, we focus on fact-checking data where claims are synthesized from multiple evidence sentences to assess the information synthesis capabilities of LLMs. We develop an automatic data construction process using simple prompt techniques on LLMs and explore several methods to improve the quality of the generated data. To evaluate the quality of the data generated by LLMs, we conduct both manual quality assessments and performance evaluations using language models. Experimental results and manual evaluations illustrate that while the quality of the generated data has significantly improved through fine-tuning techniques, LLMs still cannot match the data quality produced by humans.


Multi-Dialect Vietnamese: Task, Dataset, Baseline Models and Challenges

Van Dinh, Nguyen, Dang, Thanh Chi, Nguyen, Luan Thanh, Van Nguyen, Kiet

arXiv.org Artificial Intelligence

Vietnamese, a low-resource language, is typically categorized into three primary dialect groups that belong to Northern, Central, and Southern Vietnam. However, each province within these regions exhibits its own distinct pronunciation variations. Despite the existence of various speech recognition datasets, none of them has provided a fine-grained classification of the 63 dialects specific to individual provinces of Vietnam. To address this gap, we introduce Vietnamese Multi-Dialect (ViMD) dataset, a novel comprehensive dataset capturing the rich diversity of 63 provincial dialects spoken across Vietnam. Our dataset comprises 102.56 hours of audio, consisting of approximately 19,000 utterances, and the associated transcripts contain over 1.2 million words. To provide benchmarks and simultaneously demonstrate the challenges of our dataset, we fine-tune state-of-the-art pre-trained models for two downstream tasks: (1) Dialect identification and (2) Speech recognition. The empirical results suggest two implications including the influence of geographical factors on dialects, and the constraints of current approaches in speech recognition tasks involving multi-dialect speech data. Our dataset is available for research purposes.


Utilize Transformers for translating Wikipedia category names

Ta, Hoang-Thang, La, Quoc Thang

arXiv.org Artificial Intelligence

On Wikipedia, articles are categorized to aid readers in navigating content efficiently. The manual creation of new categories can be laborious and time-intensive. To tackle this issue, we built language models to translate Wikipedia categories from English to Vietnamese with a dataset containing 15,000 English-Vietnamese category pairs. Subsequently, small to medium-scale Transformer pre-trained models with a sequence-to-sequence architecture were fine-tuned for category translation. The experiments revealed that OPUS-MT-en-vi surpassed other models, attaining the highest performance with a BLEU score of 0.73, despite its smaller model storage. We expect our paper to be an alternative solution for translation tasks with limited computer resources.


BSRBF-KAN: A combination of B-splines and Radial Basic Functions in Kolmogorov-Arnold Networks

Ta, Hoang-Thang

arXiv.org Artificial Intelligence

A recent work of Liu et al. [1] over KANs opened a new paradigm in applying learnable activation functions as "edges" to fit training data instead of using fixed ones as "nodes" that are usually used in MLPs. The theory behind KANs relies on the Kolmogorov-Arnold representation theorem (KART), which states that a continuous function of multiple variables can be expressed as a combination of continuous functions of a single variable through additions. KANs are anticipated to bring a fresh perspective to solving issues that have been overshadowed by MLPs. With that inspiration, many scientists have flocked to develop different types of KANs based on popular polynomial and basis functions. While these works focus on univariate functions to set up KANs, none have explored their combination. Therefore, we aim to combine B-splines [2] and Radial Basis Functions [3] to build a combined KAN named BSRBF-KAN.

  Country:
  Genre: Research Report (0.50)

Self-training from Self-memory in Data-to-text Generation

Ta, Hoang-Thang

arXiv.org Artificial Intelligence

This paper introduces a novel training model, self-training from self-memory (STSM) in data-to-text generation (DTG), allowing the model to self-train on subsets, including self-memory as outputs inferred directly from the trained models and/or the new data. The quality of self-memory is validated by two models, data-to-text (D2T) and text-to-data (T2D), by two pre-defined conditions: (1) the appearance of all source values in the outputs of the D2T model and (2) the ability to convert back to source data in the outputs in the T2D model. We utilize a greedy algorithm to generate shorter D2T outputs if they contain all source values. Subsequently, we use the T2D model to confirm that these outputs can capture input relationships by demonstrating their capacity to convert text back into data. With 30% of the dataset, we can train the D2T model with a competitive performance compared to full training in the same setup. We experiment with our model on two datasets, E2E NLG and DART. STSM offers the D2T model a generalization capability from its subset memory while reducing training data volume. Ultimately, we anticipate that this paper will contribute to continual learning solutions that adapt to new training data, incorporating it as a form of self-memory in DTG tasks. The curated dataset is publicly available at: https://github.com/hoangthangta/STSM.


DepressionEmo: A novel dataset for multilabel classification of depression emotions

Rahman, Abu Bakar Siddiqur, Ta, Hoang-Thang, Najjar, Lotfollah, Azadmanesh, Azad, Gönül, Ali Saffet

arXiv.org Artificial Intelligence

Emotions are integral to human social interactions, with diverse responses elicited by various situational contexts. Particularly, the prevalence of negative emotional states has been correlated with negative outcomes for mental health, necessitating a comprehensive analysis of their occurrence and impact on individuals. In this paper, we introduce a novel dataset named DepressionEmo designed to detect 8 emotions associated with depression by 6037 examples of long Reddit user posts. This dataset was created through a majority vote over inputs by zero-shot classifications from pre-trained models and validating the quality by annotators and ChatGPT, exhibiting an acceptable level of interrater reliability between annotators. The correlation between emotions, their distribution over time, and linguistic analysis are conducted on DepressionEmo. Besides, we provide several text classification methods classified into two groups: machine learning methods such as SVM, XGBoost, and Light GBM; and deep learning methods such as BERT, GAN-BERT, and BART. The pretrained BART model, bart-base allows us to obtain the highest F1- Macro of 0.76, showing its outperformance compared to other methods evaluated in our analysis. Across all emotions, the highest F1-Macro value is achieved by suicide intent, indicating a certain value of our dataset in identifying emotions in individuals with depression symptoms through text analysis. The curated dataset is publicly available at: https://github.com/abuBakarSiddiqurRahman/DepressionEmo.


Mapping Process for the Task: Wikidata Statements to Text as Wikipedia Sentences

Ta, Hoang Thang, Gelbukha, Alexander, Sidorov, Grigori

arXiv.org Artificial Intelligence

Acknowledged as one of the most successful online cooperative projects in human society, Wikipedia has obtained rapid growth in recent years and desires continuously to expand content and disseminate knowledge values for everyone globally. The shortage of volunteers brings to Wikipedia many issues, including developing content for over 300 languages at the present. Therefore, the benefit that machines can automatically generate content to reduce human efforts on Wikipedia language projects could be considerable. In this paper, we propose our mapping process for the task of converting Wikidata statements to natural language text (WS2T) for Wikipedia projects at the sentence level. The main step is to organize statements, represented as a group of quadruples and triples, and then to map them to corresponding sentences in English Wikipedia. We evaluate the output corpus in various aspects: sentence structure analysis, noise filtering, and relationships between sentence components based on word embedding models. The results are helpful not only for the data-to-text generation task but also for other relevant works in the field.


A review of machine learning applications in wildfire science and management

Jain, Piyush, Coogan, Sean C P, Subramanian, Sriram Ganapathi, Crowley, Mark, Taylor, Steve, Flannigan, Mike D

arXiv.org Machine Learning

Artificial intelligence has been applied in wildfire science and management since the 1990s, with early applications including neural networks and expert systems. Since then the field has rapidly progressed congruently with the wide adoption of machine learning (ML) in the environmental sciences. Here, we present a scoping review of ML in wildfire science and management. Our objective is to improve awareness of ML among wildfire scientists and managers, as well as illustrate the challenging range of problems in wildfire science available to data scientists. We first present an overview of popular ML approaches used in wildfire science to date, and then review their use in wildfire science within six problem domains: 1) fuels characterization, fire detection, and mapping; 2) fire weather and climate change; 3) fire occurrence, susceptibility, and risk; 4) fire behavior prediction; 5) fire effects; and 6) fire management. We also discuss the advantages and limitations of various ML approaches and identify opportunities for future advances in wildfire science and management within a data science context. We identified 298 relevant publications, where the most frequently used ML methods included random forests, MaxEnt, artificial neural networks, decision trees, support vector machines, and genetic algorithms. There exists opportunities to apply more current ML methods (e.g., deep learning and agent based learning) in wildfire science. However, despite the ability of ML models to learn on their own, expertise in wildfire science is necessary to ensure realistic modelling of fire processes across multiple scales, while the complexity of some ML methods requires sophisticated knowledge for their application. Finally, we stress that the wildfire research and management community plays an active role in providing relevant, high quality data for use by practitioners of ML methods.